Large Margin Component Analysis

نویسندگان

  • Lorenzo Torresani
  • Kuang-chih Lee
چکیده

Metric learning has been shown to significantly improve the accuracy of k-nearest neighbor (kNN) classification. In problems involving thousands of features, distance learning algorithms cannot be used due to overfitting and high computational complexity. In such cases, previous work has relied on a two-step solution: first apply dimensionality reduction methods to the data, and then learn a metric in the resulting low-dimensional subspace. In this paper we show that better classification performance can be achieved by unifying the objectives of dimensionality reduction and metric learning. We propose a method that solves for the low-dimensional projection of the inputs, which minimizes a metric objective aimed at separating points in different classes by a large margin. This projection is defined by a significantly smaller number of parameters than metrics learned in input space, and thus our optimization reduces the risks of overfitting. Theory and results are presented for both a linear as well as a kernelized version of the algorithm. Overall, we achieve classification rates similar, and in several cases superior, to those of support vector machines.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Infinite Markov-Switching Maximum Entropy Discrimination Machines

In this paper, we present a method that combines the merits of Bayesian nonparametrics, specifically stick-breaking priors, and largemargin kernel machines in the context of sequential data classification. The proposed model employs a set of (theoretically) infinite interdependent large-margin classifiers as model components, that robustly capture local nonlinearity of complex data. The employe...

متن کامل

Investors' Perception of Bank Risk Management: Multivariate Analysis Techniques

According to the nature of their activities, banks are exposed to various types of risks. Hence, risk management is at the heart of financial institutions management. In this study, we intend to summarize the information content of bank financial statements on diverse risks faced by banks and then determine how stock markets react to bank's risk management behavior. The methodology used in this...

متن کامل

Improving Chernoff criterion for classification by using the filled function

Linear discriminant analysis is a well-known matrix-based dimensionality reduction method. It is a supervised feature extraction method used in two-class classification problems. However, it is incapable of dealing with data in which classes have unequal covariance matrices. Taking this issue, the Chernoff distance is an appropriate criterion to measure distances between distributions. In the p...

متن کامل

Maximum Margin Coresets for Active and Noise Tolerant Learning

We study the problem of learning large margin halfspaces in various settings using coresets and show that coresets are a widely applicable tool for large margin learning. A large margin coreset is a subset of the input data sufficient for approximating the true maximum margin solution. In this work, we provide a direct algorithm and analysis for constructing large margin coresets. We show vario...

متن کامل

Single Miller Capacitor Feedforward Frequency Compensation on Three Stage Amplifier for Large Capacitive Load

This paper presents the analysis of single Miller capacitor feedforward compensation scheme which is used in threestage CMOS operational amplifiers. single Miller compensation capacitor in three-stage amplifiers can significantly reduce the total capacitor value, and therefore, the overall area of the amplifiers without influencing their stability. Feed forward technique is effectively combined...

متن کامل

Corrigendum to "Cluster-based adaptive metric classification" [Neurocomputing 81 (2012) 33-40]

The components analysis family of algorithms learns transformations from constraints as explained in the survey of Yang [34]: ‘‘Relevant Components Analysis (RCA) [10] learns a global linear transformation from (a set of) equivalence constraints. The learned transformation can be used directly to compute (a) distance between any two examples. Discriminative Component Analysis (DCA) and Kernel D...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006